AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Chat optimization

# Chat optimization

GLM 4 32B 0414 4bit DWQ
MIT
This is the MLX format version of the THUDM/GLM-4-32B-0414 model, processed with 4-bit DWQ quantization, suitable for efficient inference on Apple silicon devices.
Large Language Model Supports Multiple Languages
G
mlx-community
156
4
Josiefied Qwen3 8B Abliterated V1 8bit
An optimized 8-bit quantized version of Qwen3-8B, designed for efficient inference on the MLX framework
Large Language Model
J
mlx-community
450
1
Qwen2.5 7B Instruct Quantized.w8a8
Apache-2.0
INT8 quantized version of Qwen2.5-7B-Instruct, suitable for multilingual scenarios in both commercial and research applications, optimized for memory requirements and computational throughput.
Large Language Model Safetensors English
Q
RedHatAI
412
1
H2o Danube3 4b Chat
Apache-2.0
A chat model fine-tuned by H2O.ai with 4 billion parameters, supporting text generation tasks and capable of offline operation on mobile phones.
Large Language Model Transformers English
H
h2oai
41.49k
67
Leniachat Gemma 2b V0
Apache-2.0
A 2B-parameter chat and instruction model optimized for Spanish, supporting text generation, dialogue, and Q&A tasks
Large Language Model Transformers Spanish
L
LenguajeNaturalAI
132
14
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase